# Zero-shot Cross-lingual Transfer
Trillion LLaVA 7B
Apache-2.0
Trillion-LLaVA-7B is a vision-language model (VLM) capable of understanding images, developed based on the Trillion-7B-preview foundation model.
Text-to-Image
Transformers Supports Multiple Languages

T
trillionlabs
199
8
Trillion LLaVA 7B FP16
Apache-2.0
Trillion-LLaVA-7B is a vision-language model with image understanding capabilities, trained on English visual-language instruction pairs, demonstrating exceptional cross-lingual visual reasoning abilities.
Text-to-Image
Transformers Supports Multiple Languages

T
trillionlabs
14
0
Featured Recommended AI Models